tensor factorization
- Africa > Senegal > Kolda Region > Kolda (0.05)
- North America > United States > Utah (0.04)
- North America > Canada > Quebec > Montreal (0.04)
- Europe > Switzerland > Zürich > Zürich (0.05)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- North America > United States > California > Los Angeles County > Long Beach (0.04)
- Africa > Senegal > Kolda Region > Kolda (0.05)
- North America > United States > Utah (0.04)
- North America > United States > Pennsylvania > Montgomery County (0.04)
- (2 more...)
Reviewers remark our method is intuitive and correct, and opens new directions in sparse clustering, while R1 raised
We thank you for commenting the paper is well-written and for finding a typo. Y ou suggest better baselines for comparison, citing Power k-means [37] and matrix + tensor factorization. See for instance " k-means Clustering Is Matrix Factorization" (Bauckhage 2015). We thank you for your detailed comments and careful reading of the paper. Y ou are absolutely correct that "interpretable sparsity" is overloaded here.
- North America > United States > Utah > Salt Lake County > Salt Lake City (0.04)
- Africa > Senegal > Kolda Region > Kolda (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (2 more...)
- Energy (0.46)
- Government > Regional Government (0.46)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (1.00)
- Information Technology > Data Science > Data Mining (0.69)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.46)
Duality-Induced Regularizer for Tensor Factorization Based Knowledge Graph Completion Supplementary Material
Theorem 1. Suppose that ˆ X In DB models, the commonly used p is either 1 or 2. When p = 2, DURA takes the form as the one in Equation (8) in the main text. If p = 1, we cannot expand the squared score function of the associated DB models as in Equation (4). Therefore, we choose p = 2 . 2 Table 2: Hyperparameters found by grid search. Suppose that k is the number of triplets known to be true in the knowledge graph, n is the embedding dimension of entities. That is to say, the computational complexity of weighted DURA is the same as the weighted squared Frobenius norm regularizer.
- Oceania > Australia > New South Wales > Sydney (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- Asia > China (0.04)
- Asia > Middle East > Israel > Tel Aviv District > Tel Aviv (0.04)
- Africa > Senegal > Kolda Region > Kolda (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (1.00)
- Information Technology > Data Science (0.67)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.46)
Implicit Regularization in Deep Learning May Not Be Explainable by Norms
Mathematically characterizing the implicit regularization induced by gradient-based optimization is a longstanding pursuit in the theory of deep learning. A widespread hope is that a characterization based on minimization of norms may apply, and a standard test-bed for studying this prospect is matrix factorization (matrix completion via linear neural networks).
- Asia > Middle East > Israel > Tel Aviv District > Tel Aviv (0.04)
- Africa > Senegal > Kolda Region > Kolda (0.04)
- North America > Canada (0.04)